Selective Sampling with Almost Optimal Guarantees for Learning to Rank from Pairwise Preferences
نویسندگان
چکیده
One of the practical obstacles of learning to rank from pairwise preference labels is in its (apparent) quadric sample complexity. Some heuristics have been tested for overriding this obstacle. In this workshop we will present new provable method for reducing this sample-complexity, almost reaching the informational lower bound, while suffering only negligible sacrifice of optimality. Our main results: (1) We define a novel structural property of function spaces endowed with a metric loss function that allows searching at an exponential rate. (2) We demonstrate that this structural property applies to the search space arising in the problem of learning to rank from pairwise preferences. As a result, we show that O(npoly(log n, ε−1)) adaptively sampled preferences suffice in order to obtain a ranking of ε · OPT regret, where OPT is loss of the optimal ranking. This is near-optimal in terms of information theory. (3) We show how our algorithm can be used to construct a query-efficient version of the famous SVM-rank relaxation, when the set of points are endowed with feature vectors and we are searching in the restricted space of linear permutations. (4) For feature spaces with a fixed dimension, we show an additional slight improvement in the query complexity using advanced tools from computational geometry.
منابع مشابه
Active Learning Ranking from Pairwise Preferences with Almost Optimal Query Complexity
Given a set V of n elements we wish to linearly order them using pairwise preference labels which may be non-transitive (due to irrationality or arbitrary noise). The goal is to linearly order the elements while disagreeing with as few pairwise preference labels as possible. Our performance is measured by two parameters: The number of disagreements (loss) and the query complexity (number of pai...
متن کاملDeep Unsupervised Domain Adaptation for Image Classification via Low Rank Representation Learning
Domain adaptation is a powerful technique given a wide amount of labeled data from similar attributes in different domains. In real-world applications, there is a huge number of data but almost more of them are unlabeled. It is effective in image classification where it is expensive and time-consuming to obtain adequate label data. We propose a novel method named DALRRL, which consists of deep ...
متن کاملA New Active Learning Scheme with Applications to Learning to Rank from Pairwise Preferences
We consider the statistical learning setting of active learning in which the learner chooses which examples to obtain labels for. We identify a useful general purpose structural property of such learning problems, giving rise to a query-efficient iterative procedure achieving approximately optimal loss at an exponentially fast rate, where the rate is measured in units of error per label. The ef...
متن کاملPAC Rank Elicitation through Adaptive Sampling of Stochastic Pairwise Preferences
We introduce the problem of PAC rank elicitation, which consists of sorting a given set of options based on adaptive sampling of stochastic pairwise preferences. More specifically, we assume the existence of a ranking procedure, such as Copeland’s method, that determines an underlying target order of the options. The goal is to predict a ranking that is sufficiently close to this target order w...
متن کامل"On the (Non-)existence of Convex, Calibrated Surrogate Losses for Ranking"
We study surrogate losses for learning to rank, in a framework where the rankings are induced by scores and the task is to learn the scoring function. We focus on the calibration of surrogate losses with respect to a ranking evaluation metric, where the calibration is equivalent to the guarantee that near-optimal values of the surrogate risk imply near-optimal values of the risk defined by the ...
متن کامل